teleoperation system
Hybrid-Diffusion Models: Combining Open-loop Routines with Visuomotor Diffusion Policies
Van Haastregt, Jonne, Orthmann, Bastian, Welle, Michael C., Zhang, Yuchong, Kragic, Danica
Abstract-- Despite the fact that visuomotor-based policies obtained via imitation learning demonstrate good performances in complex manipulation tasks, they usually struggle to achieve the same accuracy and speed as traditional control based methods. In this work, we introduce Hybrid-Diffusion models that combine open-loop routines with visuomotor diffusion policies. We develop T eleoperation Augmentation Primitives (T APs) that allow the operator to perform predefined routines, such as locking specific axes, moving to perching waypoints, or triggering task-specific routines seamlessly during demonstrations. Our Hybrid-Diffusion method learns to trigger such T APs during inference. All experimental videos are available on the project's website: https://hybriddiffusion. github.io/ Advances in Imitation Learning [1]-[4] have propelled autonomous manipulation capabilities to tackling complex tasks such as spreading sauce on a pizza [1], opening a capped bottle [5], inserting a hanger into a T -shirt [3], and mounting a gear on a bike [4].
ACE-F: A Cross Embodiment Foldable System with Force Feedback for Dexterous Teleoperation
Yan, Rui, Fu, Jiajian, Yang, Shiqi, Paulsen, Lars, Cheng, Xuxin, Wang, Xiaolong
Teleoperation systems are essential for efficiently collecting diverse and high-quality robot demonstration data, especially for complex, contact-rich tasks. However, current teleoperation platforms typically lack integrated force feedback, cross-embodiment generalization, and portable, user-friendly designs, limiting their practical deployment. To address these limitations, we introduce ACE-F, a cross embodiment foldable teleoperation system with integrated force feedback. Our approach leverages inverse kinematics (IK) combined with a carefully designed human-robot interface (HRI), enabling users to capture precise and high-quality demonstrations effortlessly. We further propose a generalized soft-controller pipeline integrating PD control and inverse dynamics to ensure robot safety and precise motion control across diverse robotic embodiments. Critically, to achieve cross-embodiment generalization of force feedback without additional sensors, we innovatively interpret end-effector positional deviations as virtual force signals, which enhance data collection and enable applications in imitation learning. Extensive teleoperation experiments confirm that ACE-F significantly simplifies the control of various robot embodiments, making dexterous manipulation tasks as intuitive as operating a computer mouse. The system is open-sourced at: https://acefoldable.github.io/
- North America > United States > California > San Diego County > San Diego (0.04)
- Asia > South Korea > Daegu > Daegu (0.04)
Robotic versus Human Teleoperation for Remote Ultrasound
Black, David, Salcudean, Septimiu
Abstract--Diagnostic medical ultrasound is widely used, safe, and relatively low cost but requires a high degree of expertise to acquire and interpret the images. Personnel with this expertise are often not available outside of larger cities, leading to difficult, costly travel and long wait times for rural populations. T o address this issue, tele-ultrasound techniques are being developed, including robotic teleoperation and recently human teleoperation, in which a novice user is remotely guided in a hand-overhand manner through mixed reality to perform an ultrasound exam. These methods have not been compared, and their relative strengths are unknown. Human teleoperation may be more practical than robotics for small communities due to its lower cost and complexity, but this is only relevant if the performance is comparable. This paper therefore evaluates the differences between human and robotic teleoperation, examining practical aspects such as setup time and flexibility and experimentally comparing performance metrics such as completion time, position tracking, and force consistency. It is found that human teleoperation does not lead to statistically significant differences in completion time or position accuracy, with mean differences of 1.8% and 0.5%, respectively, and provides more consistent force application despite being substantially more practical and accessible. Remote and under-resourced communities have far worse access to healthcare than larger cities [1], [2]. Ultrasound has become one of the most prevalent diagnostic imaging modalities due to its relatively low cost, non-invasive nature, and lack of radiation [3], but many communities have very limited access to qualified sonographers.
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.40)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- North America > United States > South Carolina > York County > Rock Hill (0.04)
- (3 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.68)
MR-UBi: Mixed Reality-Based Underwater Robot Arm Teleoperation System with Reaction Torque Indicator via Bilateral Control
Nishi, Kohei, Kobayashi, Masato, Uranishi, Yuki
We present a mixed reality-based underwater robot arm teleoperation system with a reaction torque indicator via bilateral control (MR-UBi). The reaction torque indicator (RTI) overlays a color and length-coded torque bar in the MR-HMD, enabling seamless integration of visual and haptic feedback during underwater robot arm teleoperation. User studies with sixteen participants compared MR-UBi against a bilateral-control baseline. MR-UBi significantly improved grasping-torque control accuracy, increasing the time within the optimal torque range and reducing both low and high grasping torque range during lift and pick-and-place tasks with objects of different stiffness. Subjective evaluations further showed higher usability (SUS) and lower workload (NASA--TLX). Overall, the results confirm that \textit{MR-UBi} enables more stable, accurate, and user-friendly underwater robot-arm teleoperation through the integration of visual and haptic feedback. For additional material, please check: https://mertcookimg.github.io/mr-ubi
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.67)
- Health & Medicine (0.93)
- Government (0.59)
Prometheus: Universal, Open-Source Mocap-Based Teleoperation System with Force Feedback for Dataset Collection in Robot Learning
Satsevich, S., Bazhenov, A., Egorov, S., Erkhov, A., Gromakov, M., Fedoseev, A., Tsetserukou, D.
This paper presents a novel teleoperation system with force feedback, utilizing consumer-grade HTC Vive Trackers 2.0. The system integrates a custom-built controller, a UR3 robotic arm, and a Robotiq gripper equipped with custom-designed fingers to ensure uniform pressure distribution on an embedded force sensor. Real-time compression force data is transmitted to the controller, enabling operators to perceive the gripping force applied to objects. Experimental results demonstrate that the system enhances task success rates and provides a low-cost solution for large-scale imitation learning data collection without compromising affordability.
- North America > United States > Texas > Travis County > Austin (0.04)
- North America > United States > Colorado > Boulder County > Boulder (0.04)
- Europe > Spain > Aragón (0.04)
- (3 more...)
AIRoA MoMa Dataset: A Large-Scale Hierarchical Dataset for Mobile Manipulation
Takanami, Ryosuke, Khrapchenkov, Petr, Morikuni, Shu, Arima, Jumpei, Takaba, Yuta, Maeda, Shunsuke, Okubo, Takuya, Sano, Genki, Sekioka, Satoshi, Kadoya, Aoi, Kambara, Motonari, Nishiura, Naoya, Suzuki, Haruto, Yoshimoto, Takanori, Sakamoto, Koya, Ono, Shinnosuke, Yang, Hu, Yashima, Daichi, Horo, Aoi, Motoda, Tomohiro, Chiyoma, Kensuke, Ito, Hiroshi, Fukuda, Koki, Goto, Akihito, Morinaga, Kazumi, Ikeda, Yuya, Kawada, Riko, Yoshikawa, Masaki, Kosuge, Norio, Noguchi, Yuki, Ota, Kei, Matsushima, Tatsuya, Iwasawa, Yusuke, Matsuo, Yutaka, Ogata, Tetsuya
As robots transition from controlled settings to unstructured human environments, building generalist agents that can reliably follow natural language instructions remains a central challenge. Progress in robust mobile manipulation requires large-scale multimodal datasets that capture contact-rich and long-horizon tasks, yet existing resources lack synchronized force-torque sensing, hierarchical annotations, and explicit failure cases. We address this gap with the AIRoA MoMa Dataset, a large-scale real-world multimodal dataset for mobile manipulation. It includes synchronized RGB images, joint states, six-axis wrist force-torque signals, and internal robot states, together with a novel two-layer annotation schema of sub-goals and primitive actions for hierarchical learning and error analysis. The initial dataset comprises 25,469 episodes (approx. 94 hours) collected with the Human Support Robot (HSR) and is fully standardized in the LeRobot v2.1 format. By uniquely integrating mobile manipulation, contact-rich interaction, and long-horizon structure, AIRoA MoMa provides a critical benchmark for advancing the next generation of Vision-Language-Action models. The first version of our dataset is now available at https://huggingface.co/datasets/airoa-org/airoa-moma .
- North America > Trinidad and Tobago > Trinidad > Arima > Arima (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
Suction Leap-Hand: Suction Cups on a Multi-fingered Hand Enable Embodied Dexterity and In-Hand Teleoperation
Zhaole, Sun, Mao, Xiaofeng, Zhu, Jihong, Zhang, Yuanlong, Fisher, Robert B.
Abstract-- Dexterous in-hand manipulation remains a foun-dational challenge in robotics, with progress often constrained by the prevailing paradigm of imitating the human hand. This anthropomorphic approach creates two critical barriers: 1) it limits robotic capabilities to tasks humans can already perform, and 2) it makes data collection for learning-based methods exceedingly difficult. Both challenges are caused by traditional force-closure which requires coordinating complex, multi-point contacts based on friction, normal force, and gravity to grasp an object. This makes teleoperated demonstrations unstable and amplifies the sim-to-real gap for reinforcement learning. In this work, we propose a paradigm shift: moving away from replicating human mechanics toward the design of novel robotic embodiments. We introduce the Suction Leap-Hand (SLeap Hand), a multi-fingered hand featuring integrated fingertip suction cups that realize a new form of suction-enabled dexterity. More importantly, this suction-based embodiment unlocks a new class of dexterous skills that are difficult or even impossible for the human hand, such as one-handed paper cutting and in-hand writing. Our work demonstrates that by moving beyond anthropomorphic constraints, novel embodiments can not only lower the barrier for collecting robust manipulation data but also enable the stable, single-handed completion of tasks that would typically require two human hands. Dexterous manipulation, the ability to reconfigure objects within a single hand, remains a grand challenge in robotics [1], [2]. The dominant paradigm for achieving this goal has been data-driven learning on anthropomorphic hands, an approach that has led to successes in grasping and reorientation [3], [4], [5].
CHILD (Controller for Humanoid Imitation and Live Demonstration): a Whole-Body Humanoid Teleoperation System
Myers, Noboru, Kwon, Obin, Yamsani, Sankalp, Kim, Joohyung
Abstract-- Recent advances in teleoperation have demonstrated robots performing complex manipulation tasks. However, existing works rarely support whole-body joint-level teleoperation for humanoid robots, limiting the diversity of tasks that can be accomplished. This work presents Controller for Humanoid Imitation and Live Demonstration (CHILD), a compact reconfigurable teleoperation system that enables joint level control over humanoid robots. CHILD fits within a standard baby carrier, allowing the operator control over all four limbs, and supports both direct joint mapping for full-body control and loco-manipulation. Adaptive force feedback is incorporated to enhance operator experience and prevent unsafe joint movements. I. INTRODUCTION Teleoperation is a commonly used technique to bridge the gap between robots' current autonomous and physical capabilities. More recently, teleoperation has become a popular method to collect demonstration data for learning-based policies.
- North America > United States > Illinois > Champaign County > Urbana (0.04)
- North America > United States > Illinois > Champaign County > Champaign (0.04)
Learning Dexterous Manipulation with Quantized Hand State
Feng, Ying, Fang, Hongjie, He, Yinong, Chen, Jingjing, Wang, Chenxi, He, Zihao, Liu, Ruonan, Lu, Cewu
Abstract-- Dexterous robotic hands enable robots to perform complex manipulations that require fine-grained control and adaptability. Achieving such manipulation is challenging because the high degrees of freedom tightly couple hand and arm motions, making learning and control difficult. Successful dexterous manipulation relies not only on precise hand motions, but also on accurate spatial positioning of the arm and coordinated arm-hand dynamics. However, most existing visuomotor policies represent arm and hand actions in a single combined space, which often causes high-dimensional hand actions to dominate the coupled action space and compromise arm control. T o address this, we propose DQ-RISE, which quantizes hand states to simplify hand motion prediction while preserving essential patterns, and applies a continuous relaxation that allows arm actions to diffuse jointly with these compact hand states. This design enables the policy to learn arm-hand coordination from data while preventing hand actions from overwhelming the action space. Experiments show that DQ-RISE achieves more balanced and efficient learning, paving the way toward structured and generalizable dexterous manipulation.
LeVR: A Modular VR Teleoperation Framework for Imitation Learning in Dexterous Manipulation
Weng, Zhengyang Kris, Elwin, Matthew L., Liu, Han
We introduce LeVR, a modular software framework designed to bridge two critical gaps in robotic imitation learning. First, it provides robust and intuitive virtual reality (VR) teleoperation for data collection using robot arms paired with dexterous hands, addressing a common limitation in existing systems. Second, it natively integrates with the powerful LeRobot imitation learning (IL) framework, enabling the use of VR-based teleoperation data and streamlining the demonstration collection process. To demonstrate LeVR, we release LeFranX, an open-source implementation for the Franka FER arm and RobotEra XHand, two widely used research platforms. LeFranX delivers a seamless, end-to-end workflow from data collection to real-world policy deployment. We validate our system by collecting a public dataset of 100 expert demonstrations and use it to successfully fine-tune state-of-the-art visuomotor policies. We provide our open-source framework, implementation, and dataset to accelerate IL research for the robotics community.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- North America > United States > Illinois > Cook County > Evanston (0.04)
- (4 more...)